Learning Bidirectional Intent Embeddings by Convolutional Deep Structured Semantic Models for Spoken Language Understanding
نویسندگان
چکیده
The recent surge of intelligent personal assistants motivates spoken language understanding of dialogue systems. Considering high-level semantics, intent embeddings can be viewed as the universal representations that help derive a more flexible intent schema to overcome the domain constraint and the genre mismatch. A convolutional deep structured semantic model (CDSSM) is applied to jointly learn the representations for human intents and associated utterances. Two sets of experiments, intent expansion and actionable item detection, are conducted to evaluate the power of the learned intent embeddings. The representations bridge the semantic relation between seen and unseen intents for intent expansion, and connect intents from different genres for actionable item detection. The discussion and analysis of experiments provide a future direction for reducing human effort of data annotation and eliminating domain and genre constraints for spoken language understanding.
منابع مشابه
Unsupervised Learning and Modeling of Knowledge and Intent for Spoken Dialogue Systems
Various smart devices (smartphone, smart-TV, in-car navigating system, etc.) are incorporating spoken language interfaces, as known as spoken dialogue systems (SDS), to help users finish tasks more efficiently. The key role in a successful SDS is a spoken language understanding (SLU) component; in order to capture language variation from dialogue participants, the SLU component must create a ma...
متن کاملExploiting Sentence and Context Representations in Deep Neural Models for Spoken Language Understanding
This paper presents a deep learning architecture for the semantic decoder component of a Statistical Spoken Dialogue System. In a slot-filling dialogue, the semantic decoder predicts the dialogue act and a set of slot-value pairs from a set of n-best hypotheses returned by the Automatic Speech Recognition. Most current models for spoken language understanding assume (i) word-aligned semantic an...
متن کاملMulti-Task Learning for parsing the Alexa Meaning Representation Language
The Alexa Meaning Representation Language (AMRL) is a compositional graph-based semantic representation that includes fine-grained types, properties, actions, and roles and can represent a wide variety of spoken language. AMRL increases the ability of virtual assistants to represent more complex requests, including logical and conditional statements as well as ones with nested clauses. Due to t...
متن کاملDeep contextual language understanding in spoken dialogue systems
We describe a unified multi-turn multi-task spoken language understanding (SLU) solution capable of handling multiple context sensitive classification (intent determination) and sequence labeling (slot filling) tasks simultaneously. The proposed architecture is based on recurrent convolutional neural networks (RCNN) with shared feature layers and globally normalized sequence modeling components...
متن کاملMulti-Domain Joint Semantic Frame Parsing Using Bi-Directional RNN-LSTM
Sequence-to-sequence deep learning has recently emerged as a new paradigm in supervised learning for spoken language understanding. However, most of the previous studies explored this framework for building single domain models for each task, such as slot filling or domain classification, comparing deep learning based approaches with conventional ones like conditional random fields. This paper ...
متن کامل